Severity of Local Maxima for the EM Algorithm: Experiences with Hierarchical Latent Class Models

نویسندگان

  • Yi Wang
  • Nevin Lianwen Zhang
چکیده

It is common knowledge that the EM algorithm can be trapped at local maxima and consequently fails to reach global maxima. We empirically investigate the severity of this problem in the context of hierarchical latent class (HLC) models. Our experiments were run on HLC models where dependency between neighboring variables is strong. (The reason for focusing on this class of models will be made clear in the main text.) We first ran EM from randomly generated single starting points, and observed that (1) the probability of hitting global maxima is generally high, (2) it increases with the strength of dependency and sample sizes, and (3) it decreases with the amount of extreme probability values. We also observed that, at high dependence strength levels, local maxima are far apart from global ones in terms of likelihoods. Those imply that local maxima can be reliably avoided by running EM from a few starting points and hence are not a serious issue. This is confirmed by our second set of experiments.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Prague , Czech Republic September 12 − 15 , 2006 Probabilistic Graphical Models

ion and Refinement for Solving Continuous Markov Decision Processes. . . . . . . . . . . . . . .263Alberto Reyes, Pablo Ibargüengoytia, L. Enrique Sucar, and Eduardo Morales Bayesian Model Averaging of TAN Models for Clustering. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .271Guzmán Santafé, Jose A. Lozano, and Pedro Larrañaga Dynamic Weighting A* Search-based ...

متن کامل

Structural EM for Hierarchical Latent Class Models

Hierarchical latent class (HLC) models are tree-structured Bayesian networks where leaf nodes are observed while internal nodes are not. This paper is concerned with the problem of learning HLC models from data. We apply the idea of structural EM to a hill-climbing algorithm for this task described in an accompanying paper (Zhang et al. 2003) and show empirically that the improved algorithm can...

متن کامل

Learning of Latent Class Models by Splitting and Merging Components

A problem in learning latent class models (also known as naive Bayes models with a hidden class variable) is that local maximum parameters are often found. This leads not only to suboptimal parameters, but also to a wrong number of classes (components) for a hidden variable. The standard solution of having many random starting points for the EM algorithm is often too expensive computationally. ...

متن کامل

A Note on the Expectation-Maximization (EM) Algorithm

The Expectation-Maximization (EM) algorithm is a general algorithm for maximum-likelihood estimation where the data are “incomplete” or the likelihood function involves latent variables. Note that the notion of “incomplete data” and “latent variables” are related: when we have a latent variable, we may regard our data as being incomplete since we do not observe values of the latent variables; s...

متن کامل

An EM algorithm for the estimation of parametric and nonparametric hierarchical nonlinear models

It is shown how to implement an EM algorithm for maximum likelihood estimation of hierarchical nonlinear models for data sets consisting of more than two levels of nesting. This upward–downward algorithm makes use of the conditional independence assumptions implied by the hierarchical model. It cannot only be used for the estimation of models with a parametric specification of the random effect...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2006